Online Learning and Mistake Bounds for Finite Hypothesis Classes
نویسنده
چکیده
In this lecture we describe a different model of learning which is called online learning. Online learning takes place in a sequence of consecutive rounds. To demonstrate the online learning model, consider again the papaya tasting problem. On each online round, the learner first receives an instance (the learner buys a papaya and knows its shape and color, which form the instance). Then, the learner is required to predict a label (is the papaya tasty?). At the end of the round, the learner gets the correct label (he tastes the papaya and then knows if it’s tasty or not). Finally, the learner uses this information to improve his future predictions. Previously, we used the batch learning model in which we first use a batch of training examples to learn a hypothesis and only when learning is completed the learned hypothesis is tested. In our papayas learning problem, we should first buy bunch of papayas and taste them all. Then, we use all of this information to learn a prediction rule that determines the taste of new papayas. In contrast, in online learning there is no separation between a training phase and a test phase. The same sequence of examples is used both for training and testing and the distinguish between train and test is through time. In our papaya problem, each time we buy a papaya, it is first considered a test example since we should predict if it’s going to taste good. But, after we take a byte from the papaya, we know the true label, and the same papaya becomes a training example that can help us improve our prediction mechanism for future papayas. The goal of the online learner is simply to make few prediction mistakes. By now, the reader should know that there are no free lunches – we must have some prior knowledge on the problem in order to be able to make accurate predictions. As in previous lectures, we encode our prior knowledge on the problem using some representation of the instances (e.g. shape and color) and by assuming that there is a class of hypotheses, H = {h : X → Y}, and on each online round the learner uses a hypothesis from H to make his prediction. To simplify our presentation, we start the lecture by describing online learning algorithms for the case of a finite hypothesis class.
منابع مشابه
Online Learning Meets Optimization in the Dual
We describe a novel framework for the design and analysis of online learning algorithms. Our framework is based on a new perspective on relative mistake bounds by viewing the number of mistakes of an online learning algorithm as a lower bound for an optimization problem. This interpretation of a mistake bound draws a connection between online learning and optimization through the theory of dual...
متن کاملA Multi-class Linear Learning Algorithm Related to Winnow
In this paper, we present Committee, a new multi-class learning algorithm related to the Winnow family of algorithms. Committee is an algorithm for combining the predictions of a set of sub-experts in the online mistake-bounded model of learning. A sub-expert is a special type of attribute that predicts with a distribution over a finite number of classes. Committee learns a linear function of s...
متن کاملA Multi-class Linear Learning Algorithm Related to Winnow with Proof
In this paper, we present Committee, a new multi-class learning algorithm related to the Winnow family of algorithms. Committee is an algorithm for combining the predictions of a set of sub-experts in the online mistake-bounded model of learning. A sub-expert is a special type of attribute that predicts with a distribution over a finite number of classes. Committee learns a linear function of s...
متن کاملCovering Number Bounds of Certain Regularized Linear Function Classes
Recently, sample complexity bounds have been derived for problems involving linear functions such as neural networks and support vector machines. In many of these theoretical studies, the concept of covering numbers played an important role. It is thus useful to study covering numbers for linear function classes. In this paper, we investigate two closely related methods to derive upper bounds o...
متن کاملOnline Learning versus Ooine Learning
x Abstract. We present an oo-line variant of the mistake-bound model of learning. Just like in the well studied on-line model, a learner in the oo-line model has to learn an unknown concept from a sequence of elements of the instance space on which he makes \guess and test" trials. In both models, the aim of the learner is to make as few mistakes as possible. The diierence between the models is...
متن کامل